18 research outputs found
Multimodal Manoeuvre and Trajectory Prediction for Autonomous Vehicles Using Transformer Networks
Predicting the behaviour (i.e. manoeuvre/trajectory) of other road users,
including vehicles, is critical for the safe and efficient operation of
autonomous vehicles (AVs), a.k.a. automated driving systems (ADSs). Due to the
uncertain future behaviour of vehicles, multiple future behaviour modes are
often plausible for a vehicle in a given driving scene. Therefore, multimodal
prediction can provide richer information than single-mode prediction enabling
AVs to perform a better risk assessment. To this end, we propose a novel
multimodal prediction framework that can predict multiple plausible behaviour
modes and their likelihoods. The proposed framework includes a bespoke problem
formulation for manoeuvre prediction, a novel transformer-based prediction
model, and a tailored training method for multimodal manoeuvre and trajectory
prediction. The performance of the framework is evaluated using two public
benchmark highway driving datasets, namely NGSIM and highD. The results show
that the proposed framework outperforms the state-of-the-art multimodal methods
in the literature in terms of prediction error and is capable of predicting
plausible manoeuvre and trajectory modes.Comment: 8 pages, 3 figures, submitted to IEEE RA
Early Lane Change Prediction for Automated Driving Systems Using Multi-Task Attention-based Convolutional Neural Networks
Lane change (LC) is one of the safety-critical manoeuvres in highway driving
according to various road accident records. Thus, reliably predicting such
manoeuvre in advance is critical for the safe and comfortable operation of
automated driving systems. The majority of previous studies rely on detecting a
manoeuvre that has been already started, rather than predicting the manoeuvre
in advance. Furthermore, most of the previous works do not estimate the key
timings of the manoeuvre (e.g., crossing time), which can actually yield more
useful information for the decision making in the ego vehicle. To address these
shortcomings, this paper proposes a novel multi-task model to simultaneously
estimate the likelihood of LC manoeuvres and the time-to-lane-change (TTLC). In
both tasks, an attention-based convolutional neural network (CNN) is used as a
shared feature extractor from a bird's eye view representation of the driving
environment. The spatial attention used in the CNN model improves the feature
extraction process by focusing on the most relevant areas of the surrounding
environment. In addition, two novel curriculum learning schemes are employed to
train the proposed approach. The extensive evaluation and comparative analysis
of the proposed method in existing benchmark datasets show that the proposed
method outperforms state-of-the-art LC prediction models, particularly
considering long-term prediction performance.Comment: 13 pages, 11 figure
Fast and robust registration of partially overlapping point clouds
Real-time registration of partially overlapping point clouds has emerging applications in cooperative perception for autonomous vehicles and multi-agent SLAM. The relative translation between point clouds in these applications is higher than in traditional SLAM and odometry applications, which challenges the identification of correspondences and a successful registration. In this paper, we propose a novel registration method for partially overlapping point clouds where correspondences are learned using an efficient point-wise feature encoder, and refined using a graph-based attention network. This attention network exploits geometrical relationships between key points to improve the matching in point clouds with low overlap. At inference time, the relative pose transformation is obtained by robustly fitting the correspondences through sample consensus. The evaluation is performed on the KITTI dataset and a novel synthetic dataset including low-overlapping point clouds with displacements of up to 30m. The proposed method achieves on-par performance with state-of-the-art methods on the KITTI dataset, and outperforms existing methods for low overlapping point clouds. Additionally, the proposed method achieves significantly faster inference times, as low as 410ms, between 5 and 35 times faster than competing methods. Our code and dataset will be available at https://github.com/eduardohenriquearnold/fastreg
Deep Learning-based Vehicle Behaviour Prediction For Autonomous Driving Applications: A Review
Behaviour prediction function of an autonomous vehicle predicts the future
states of the nearby vehicles based on the current and past observations of the
surrounding environment. This helps enhance their awareness of the imminent
hazards. However, conventional behaviour prediction solutions are applicable in
simple driving scenarios that require short prediction horizons. Most recently,
deep learning-based approaches have become popular due to their superior
performance in more complex environments compared to the conventional
approaches. Motivated by this increased popularity, we provide a comprehensive
review of the state-of-the-art of deep learning-based approaches for vehicle
behaviour prediction in this paper. We firstly give an overview of the generic
problem of vehicle behaviour prediction and discuss its challenges, followed by
classification and review of the most recent deep learning-based solutions
based on three criteria: input representation, output type, and prediction
method. The paper also discusses the performance of several well-known
solutions, identifies the research gaps in the literature and outlines
potential new research directions
A Novel Deep Neural Network for Trajectory Prediction in Automated Vehicles Using Velocity Vector Field
Anticipating the motion of other road users is crucial for automated driving
systems (ADS), as it enables safe and informed downstream decision-making and
motion planning. Unfortunately, contemporary learning-based approaches for
motion prediction exhibit significant performance degradation as the prediction
horizon increases or the observation window decreases. This paper proposes a
novel technique for trajectory prediction that combines a data-driven
learning-based method with a velocity vector field (VVF) generated from a
nature-inspired concept, i.e., fluid flow dynamics. In this work, the vector
field is incorporated as an additional input to a convolutional-recurrent deep
neural network to help predict the most likely future trajectories given a
sequence of bird's eye view scene representations. The performance of the
proposed model is compared with state-of-the-art methods on the HighD dataset
demonstrating that the VVF inclusion improves the prediction accuracy for both
short and long-term (5~sec) time horizons. It is also shown that the accuracy
remains consistent with decreasing observation windows which alleviates the
requirement of a long history of past observations for accurate trajectory
prediction. Source codes are available at:
https://github.com/Amir-Samadi/VVF-TP.Comment: This paper has been accepted and nominated as the best student paper
at the 26th IEEE International Conference on Intelligent Transportation
Systems (ITSC 2023
Deep learning-based vehicle behaviour prediction for autonomous driving applications : a review
Behaviour prediction function of an autonomous vehicle predicts the future states of the nearby vehicles based on the current and past observations of the surrounding environment. This helps enhance their awareness of the imminent hazards. However, conventional behavior prediction solutions are applicable in simple driving scenarios that require short prediction horizons. Most recently, deep learning-based approaches have become popular due to their promising performance in more complex environments compared to the conventional approaches. Motivated by this increased popularity, we provide a comprehensive review of the state-of-the-art of deep learning-based approaches for vehicle behavior prediction in this article. We firstly give an overview of the generic problem of vehicle behavior prediction and discuss its challenges, followed by classification and review of the most recent deep learning-based solutions based on three criteria: input representation, output type, and prediction method. The article also discusses the performance of several well-known solutions, identifies the research gaps in the literature and outlines potential new research directions
A comparative study of ego-centric and cooperative perception for lane change prediction in highway driving scenarios
Automated vehicles are required to predict the future manoeuvres, e.g., lane change, of their nearby vehicles to operate safely and efficiently. A Manoeuvre prediction algorithm estimates the likelihood of a target vehicle's next manoeuvre using the motion history of the target vehicle and its surrounding traffic. Several existing studies assume full observability of the surrounding traffic by utilising trajectory datasets collected by wide and top-down view infrastructure cameras. However, in practice, automated vehicles observe the driving environment using ego-centric perception sensors (i.e., onboard lidar or camera) which have limited sensing range and are subject to occlusion.
To overcome these limitations, automated vehicles can cooperate in observing the environment by sharing their perception data through V2V communication. This paper analyses the impact of ego-centric and cooperative perception on the lane change prediction problem. To this end, we propose two perception models used to generate egocentric and cooperative perception dataset variants from a widely used top-down view trajectory dataset. The evaluation results show that egocentric perception decreases the performance of our long-term lane change prediction model by 4% compared to the full observability mode in the original dataset. Furthermore, the results suggest that using cooperative perception with 20% penetration rate of automated vehicles significantly mitigate the performance loss caused by the limitations of egocentric perception
Multimodal manoeuvre and trajectory prediction for automated driving on highways using transformer networks
Predicting the behaviour (i.e., manoeuvre/trajectory) of other road users, including vehicles, is critical for the safe and efficient operation of autonomous vehicles (AVs), a.k.a., automated driving systems (ADSs). Due to the uncertain future behaviour of vehicles, multiple future behaviour modes are often plausible for a vehicle in a given driving scene. Therefore, multimodal prediction can provide richer information than single-mode prediction, enabling AVs to perform a better risk assessment. To this end, we propose a novel multimodal prediction framework that can predict multiple plausible behaviour modes and their likelihoods. The proposed framework includes a bespoke problem formulation for manoeuvre prediction, a novel transformer-based prediction model, and a tailored training method for multimodal manoeuvre and trajectory prediction. The performance of the framework is evaluated using three public highway driving datasets, namely NGSIM, highD, and exiD. The results show that our framework outperforms the state-of-the-art multimodal methods in terms of prediction error and is capable of predicting plausible manoeuvre and trajectory modes
A novel deep neural network for trajectory prediction in automated vehicles using velocity vector field
Anticipating the motion of other road users is crucial for automated driving systems (ADS), as it enables safe and informed downstream decision-making and motion planning. Unfortunately, contemporary learning-based approaches for motion prediction exhibit significant performance degradation as the prediction horizon increases or the observation window decreases. This paper proposes a novel technique for trajectory prediction that combines a data-driven learningbased method with a velocity vector field (VVF) generated from a nature-inspired concept, i.e., fluid flow dynamics. In this work, the vector field is incorporated as an additional input to a convolutional-recurrent deep neural network to help predict the most likely future trajectories given a sequence of bird’s eye view scene representations. The performance of the proposed model is compared with state-of-the-art methods on the highD dataset demonstrating that the VVF inclusion improves the prediction accuracy for both short and longterm (5 sec) time horizons. It is also shown that the accuracy remains consistent with decreasing observation windows which alleviates the requirement of a long history of past observations for accurate trajectory prediction